Skip to main content

Discrete Random Variables: Quantitative Insights into Random Phenomena

In probability theory and statistics, the concept of a random variable serves as a mathematical abstraction to quantify randomness. While random variables can be either continuous or discrete, this section specifically focuses on discrete random variables, breaking down their mathematical formulation, characteristics, and applications.

Definition

A discrete random variable XX is a function that maps each outcome in a sample space SS to a real number:

X:SRX: S \rightarrow \mathbb{R}

In other words, a discrete random variable assigns a numerical value to each possible outcome of a random experiment.

To elucidate these concepts, let's consider a basic example: rolling a fair six-sided die. The sample space SS is {1,2,3,4,5,6}\{1, 2, 3, 4, 5, 6\}.

Probability Mass Function (PMF)

The PMF, denoted as pX(x)p_X(x), assigns probabilities to the values that XX can take. For our die roll:

pX(x)=16,x{1,2,3,4,5,6}p_X(x) = \frac{1}{6}, \quad x \in \{1, 2, 3, 4, 5, 6\}

Cumulative Distribution Function (CDF)

The CDF FX(x)F_X(x) is the probability that XX takes on a value less than or equal to xx:

FX(x)=txpX(t)F_X(x) = \sum_{t \leq x} p_X(t)

For our die roll, FX(3)=pX(1)+pX(2)+pX(3)=16+16+16=12F_X(3) = p_X(1) + p_X(2) + p_X(3) = \frac{1}{6} + \frac{1}{6} + \frac{1}{6} = \frac{1}{2}.

Expectation

The expectation E[X]\mathbb{E}[X] is the "average" outcome, calculated as:

E[X]=xSxpX(x)=116+216++616=3.5\mathbb{E}[X] = \sum_{x \in S} x \cdot p_X(x) = 1 \cdot \frac{1}{6} + 2 \cdot \frac{1}{6} + \ldots + 6 \cdot \frac{1}{6} = 3.5

Variance

The variance Var(X)\text{Var}(X) measures how dispersed the values are around the mean:

Var(X)=xS(xE[X])2pX(x)=((13.5)2++(63.5)2)16=2.92\text{Var}(X) = \sum_{x \in S} (x - \mathbb{E}[X])^2 \cdot p_X(x) = \left( (1 - 3.5)^2 + \ldots + (6 - 3.5)^2 \right) \frac{1}{6} = 2.92

Applications and Implications

Discrete random variables often model countable phenomena like the number of heads in a sequence of coin flips, or the number of defective items in a batch. They are indispensable in fields like machine learning for classification tasks, economics for modeling consumer choice, and operations research for optimizing discrete systems.

Bernoulli and Binomial Distributions

These are classic examples of discrete random variables. A Bernoulli distribution models a single trial (success/failure), whereas a binomial distribution extends this to nn independent trials.

Bernoulli Distribution

A Bernoulli distribution models a single trial of a random experiment that results in a success with probability pp or failure with probability 1p1-p.

Definition and PMF

The probability mass function (PMF) of a Bernoulli distribution XX is defined as:

pX(x)={p,if x=11p,if x=0p_X(x) = \begin{cases} p, & \text{if } x = 1 \\ 1-p, & \text{if } x = 0 \end{cases}

Expectation and Variance

The expectation E[X]\mathbb{E}[X] and variance Var(X)\text{Var}(X) are:

E[X]=1p+0(1p)=p,Var(X)=(1p)2(1p)+(1p)2p=p(1p)\mathbb{E}[X] = 1 \cdot p + 0 \cdot (1-p) = p, \quad \text{Var}(X) = (1 - p)^2 \cdot (1-p) + (1 - p)^2 \cdot p = p(1-p)

Binomial Distribution

When nn independent Bernoulli trials are conducted, the number of successes is modeled by a Binomial distribution.

Definition and PMF

For nn independent Bernoulli trials each with success probability pp, the PMF of a Binomial distribution XX is:

pX(k)=(nk)pk(1p)(nk),k{0,1,2,,n}p_X(k) = \binom{n}{k} p^k (1-p)^{(n-k)}, \quad k \in \{0,1,2,\ldots,n\}

where (nk)\binom{n}{k} is the number of combinations of nn objects taken kk at a time.

Expectation and Variance

For a Binomial distribution, the expectation and variance are:

E[X]=np,Var(X)=np(1p)\mathbb{E}[X] = n \cdot p, \quad \text{Var}(X) = n \cdot p \cdot (1-p)

Example: Coin Flipping

Suppose we flip a fair coin n=5n = 5 times. Each flip is a Bernoulli trial with p=0.5p = 0.5.

The PMF of the Binomial distribution XX for this case is:

pX(k)=(5k)(12)k(12)(5k),k{0,1,2,3,4,5}p_X(k) = \binom{5}{k} \left( \frac{1}{2} \right)^k \left( \frac{1}{2} \right)^{(5-k)}, \quad k \in \{0,1,2,3,4,5\}

The expectation and variance would be:

E[X]=5×0.5=2.5,Var(X)=5×0.5×0.5=1.25\mathbb{E}[X] = 5 \times 0.5 = 2.5, \quad \text{Var}(X) = 5 \times 0.5 \times 0.5 = 1.25

Conclusions

Discrete random variables provide a robust mathematical framework for quantifying and reasoning about uncertainty in countable spaces. Through their PMFs, CDFs, and moments (like expectation and variance), they offer a plethora of avenues for both theoretical inquiry and empirical analysis.

FIGURE TO ADD LATER here would be a plot of a PMF and CDF for a typical discrete random variable, illustrating how probabilities are distributed over the sample space.